A Block Successive Upper Bound Minimization Method of Multipliers for Linearly Constrained Convex Optimization
نویسندگان
چکیده
Consider the problem of minimizing the sum of a smooth convex function and a separable nonsmooth convex function subject to linear coupling constraints. Problems of this form arise in many contemporary applications including signal processing, wireless networking and smart grid provisioning. Motivated by the huge size of these applications, we propose a new class of first order primal-dual algorithms called the block successive upper-bound minimization method of multipliers (BSUM-M) to solve this family of problems. The BSUM-M updates the primal variable blocks successively by minimizing locally tight upper-bounds of the augmented Lagrangian of the original problem, followed by a gradient type update for the dual variable in closed form. We show that under certain regularity conditions, and when the primal block variables are updated in either a deterministic or a random fashion, the BSUM-M converges to the set of optimal solutions. Moreover, in the absence of linear constraints, we show that the BSUM-M, which reduces to the block successive upper-bound minimization (BSUM [1]) method, is capable of linear convergence without strong convexity.
منابع مشابه
On the Convergence Properties of a Majorized Alternating Direction Method of Multipliers for Linearly Constrained Convex Optimization Problems with Coupled Objective Functions
In this paper, we establish the convergence properties for a majorized alternating direction method of multipliers for linearly constrained convex optimization problems,whose objectives contain coupled functions.Our convergence analysis relies on the generalized Mean-Value Theorem, which plays an important role to properly control the cross terms due to the presence of coupled objective functio...
متن کاملBlock-wise ADMM with a Relaxation Factor for Multiple-block Convex Programming
It has been shown that the alternating direction method of multipliers (ADMM) is not necessarily convergent when it is directly extended to a multiple-block linearly constrained convex minimization model with an objective function in the sum of more than two functions without coupled variables. Most recently, we proposed the block-wise ADMM, which was obtained by regrouping the variables and fu...
متن کاملLinearized Alternating Direction Method of Multipliers for Constrained Nonconvex Regularized Optimization
In this paper, we consider a wide class of constrained nonconvex regularized minimization problems, where the constraints are linearly constraints. It was reported in the literature that nonconvex regularization usually yields a solution with more desirable sparse structural properties beyond convex ones. However, it is not easy to obtain the proximal mapping associated with nonconvex regulariz...
متن کاملOn the Convergence of Multi-Block Alternating Direction Method of Multipliers and Block Coordinate Descent Method
The paper answers several open questions of the alternating direction method of multipliers (ADMM) and the block coordinate descent (BCD) method that are now wildly used to solve large scale convex optimization problems in many fields. For ADMM, it is still lack of theoretical understanding of the algorithm when the objective function is not separable across the variables. In this paper, we ana...
متن کاملOn the duality of quadratic minimization problems using pseudo inverses
In this paper we consider the minimization of a positive semidefinite quadratic form, having a singular corresponding matrix $H$. We state the dual formulation of the original problem and treat both problems only using the vectors $x in mathcal{N}(H)^perp$ instead of the classical approach of convex optimization techniques such as the null space method. Given this approach and based on t...
متن کامل